The Functional Definition (Definition 2.1.1)
A random variable $X$ is a function $X: S \to R^1$ that assigns a real number $X(s)$ to every possible outcome $s$ in the sample space $S$. Refer to Figure 2.1.1 for the visual mapping of this process.
To bridge set theory and arithmetic, we define the indicator function of an event $A$:
$$I_A(s) = \begin{cases} 1 & s \in A \\ 0 & s \notin A \end{cases}$$
This transforms the occurrence of an event into a binary numerical signal.
Defining Distributions (Definition 2.2.1)
The "distribution" of $X$ is the collection of probabilities $P(X \in B)$ for subsets $B \subseteq R^1$. Strictly speaking, it is required that $B$ be a Borel subset, which is a technical restriction from measure theory. However, any subset we can practically define is a Borel subset.
Limits and Continuity of Probability
To ensure our functions behave predictably in infinite contexts, we rely on the axioms established in Theorems 1.3.4 and 1.6.1:
- Countable Additivity (1.7.1): $P(A_1 \cup A_2 \cup \cdots) = \sum P(B_n)$, where $B_n$ are disjoint versions of $A_n$.
- Continuity of Probability (1.7.2): If a sequence of events $\{A_n\} \nearrow A$, then $\lim_{n \to \infty} P(A_n) = P(A)$.
We want to prove that for any sequence of events $A_1, A_2, \dots$ (not necessarily disjoint):
$$P(A_1 \cup A_2 \cup \cdots) \leq P(A_1) + P(A_2) + \cdots$$
This is known as Boole's Inequality and is fundamental to bounding probabilities in complex systems.